翻訳と辞書
Words near each other
・ Non-surgical liposuction
・ Non-surgical rhinoplasty
・ Non-synchronous transmission
・ Non-system opposition
・ Non-tariff barriers to trade
・ Non-tax revenue
・ Non-thermal microwave effect
・ Non-timber forest product
・ Non-topological soliton
・ Non-towered airport
・ Non-traditional student
・ Non-trophic networks
・ Non-tropic hormone
・ Non-U.S. recipients of U.S. gallantry awards
・ Non-uniform discrete Fourier transform
Non-uniform memory access
・ Non-uniform rational B-spline
・ Non-use value
・ Non-vascular plant
・ Non-vegetarian
・ Non-verbal leakage
・ Non-Verbal Reasoning
・ Non-violation nullification of benefits
・ Non-Violence (sculpture)
・ Non-violent abortion protests
・ Non-virtual interface pattern
・ Non-Volatile File System
・ Non-volatile memory
・ Non-volatile random-access memory
・ Non-volcanic passive margins


Dictionary Lists
翻訳と辞書 辞書検索 [ 開発暫定版 ]
スポンサード リンク

Non-uniform memory access : ウィキペディア英語版
Non-uniform memory access

Non-uniform memory access (NUMA) is a computer memory design used in multiprocessing, where the memory access time depends on the memory location relative to the processor. Under NUMA, a processor can access its own local memory faster than non-local memory (memory local to another processor or memory shared between processors). The benefits of NUMA are limited to particular workloads, notably on servers where the data are often associated strongly with certain tasks or users.
NUMA architectures logically follow in scaling from symmetric multiprocessing (SMP) architectures. They were developed commercially during the 1990s by Burroughs (later Unisys), Convex Computer (later Hewlett-Packard), Honeywell Information Systems Italy (HISI) (later Groupe Bull), Silicon Graphics (later Silicon Graphics International), Sequent Computer Systems (later IBM), Data General (later EMC), and Digital (later Compaq, now HP). Techniques developed by these companies later featured in a variety of Unix-like operating systems, and to an extent in Windows NT.
The first commercial implementation of a NUMA-based Unix system was the Symmetrical Multi Processing XPS-100 family of servers, designed by Dan Gielan of VAST Corporation for Honeywell Information Systems Italy.
==Basic concept==

Modern CPUs operate considerably faster than the main memory they use. In the early days of computing and data processing, the CPU generally ran slower than its own memory. The performance lines of processors and memory crossed in the 1960s with the advent of the first supercomputers. Since then, CPUs increasingly have found themselves "starved for data" and having to stall while waiting for data to arrive from memory. Many supercomputer designs of the 1980s and 1990s focused on providing high-speed memory access as opposed to faster processors, allowing the computers to work on large data sets at speeds other systems could not approach.
Limiting the number of memory accesses provided the key to extracting high performance from a modern computer. For commodity processors, this meant installing an ever-increasing amount of high-speed cache memory and using increasingly sophisticated algorithms to avoid cache misses. But the dramatic increase in size of the operating systems and of the applications run on them has generally overwhelmed these cache-processing improvements. Multi-processor systems without NUMA make the problem considerably worse. Now a system can starve several processors at the same time, notably because only one processor can access the computer's memory at a time.
NUMA attempts to address this problem by providing separate memory for each processor, avoiding the performance hit when several processors attempt to address the same memory. For problems involving spread data (common for servers and similar applications), NUMA can improve the performance over a single shared memory by a factor of roughly the number of processors (or separate memory banks). Another approach to addressing this problem, utilized mainly by non-NUMA systems, is the multi-channel memory architecture; multiple memory channels are increasing the number of simultaneous memory accesses.〔

Of course, not all data ends up confined to a single task, which means that more than one processor may require the same data. To handle these cases, NUMA systems include additional hardware or software to move data between memory banks. This operation slows the processors attached to those banks, so the overall speed increase due to NUMA depends heavily on the nature of the running tasks.〔
Intel announced NUMA compatibility for its x86 and Itanium servers in late 2007 with its Nehalem and Tukwila CPUs.〔Intel Corp. (2008). Intel QuickPath Architecture (paper ). Retrieved from http://www.intel.com/pressroom/archive/reference/whitepaper_QuickPath.pdf〕 Both CPU families share a common chipset; the interconnection is called Intel Quick Path Interconnect (QPI).〔Intel Corporation. (September 18th, 2007). Gelsinger Speaks To Intel And High-Tech Industry's Rapid Technology Caden(release ). Retrieved from http://www.intel.com/pressroom/archive/releases/2007/20070918corp_b.htm〕 AMD implemented NUMA with its Opteron processor (2003), using HyperTransport.

抄文引用元・出典: フリー百科事典『 ウィキペディア(Wikipedia)
ウィキペディアで「Non-uniform memory access」の詳細全文を読む



スポンサード リンク
翻訳と辞書 : 翻訳のためのインターネットリソース

Copyright(C) kotoba.ne.jp 1997-2016. All Rights Reserved.